Goto

Collaborating Authors

 use case


Drink Whole Milk, Eat Red Meat, and Use ChatGPT

The Atlantic - Technology

Robert F. Kennedy Jr. is an AI guy. Last week, during a stop in Nashville on his Take Back Your Health tour, the Health and Human Services secretary brought up the technology between condemning ultra-processed foods and urging Americans to eat protein. "My agency is now leading the federal government in driving AI into all of our activities," he declared. An army of bots, Kennedy said, will transform medicine, eliminate fraud, and put a virtual doctor in everyone's pocket. RFK Jr. has talked up the promise of infusing his department with AI for months.


I Have Fallen in Love With Open Earbuds (and You Should Too)

WIRED

From jogging and cycling to multi-tasking or puttering around the house, open earbuds are an excellent way to jam out in the real world. If you've done any wireless earbuds shopping lately, you've likely noticed a new design category cropping up everywhere. They're called open earbuds (or open-ear buds, depending on the brand), and just about every audio brand has a pair (or three). They come in a slew of styles, but most either loop around your ears like older Beats buds, or clip on like funky-futuristic earrings. Whatever the style, they're designed to deliver satisfying sound while keeping your ear canals open to the sounds of the world around you.


The crucial first step for designing a successful enterprise AI system

MIT Technology Review

How to identify the first iconic use case for an enterprise AI transformation. Many organizations rushed into generative AI, only to see pilots fail to deliver value . Now, companies want measurable outcomes--but how do you design for success? At Mistral AI, we partner with global industry leaders to co-design tailored AI solutions that solve their most difficult problems. Whether it's increasing CX productivity with Cisco, building a more intelligent car with Stellantis, or accelerating product innovation with ASML, we start with open frontier models and customize AI systems to deliver impact for each company's unique challenges and goals. Our methodology starts by identifying an iconic use case, the foundation for AI transformation that sets the blueprint for future AI solutions.



ALI-Agent: Assessing LLMs' Alignment with Human Values via Agent-based Evaluation

Neural Information Processing Systems

To mitigate these risks, current evaluation benchmarks predominantly employ expert-designed contextual scenarios to assess how well LLMs align with human values. However, the labor-intensive nature of these benchmarks limits their test scope, hindering their ability to generalize to the extensive variety of open-world use cases and identify rare but crucial long-tail risks. Additionally, these static tests fail to adapt to the rapid evolution of LLMs, making it hard to evaluate timely alignment issues. To address these challenges, we propose ALI-Agent, an evaluation framework that leverages the autonomous abilities of LLM-powered agents to conduct in-depth and adaptive alignment assessments. ALI-Agent operates through two principal stages: Emulation and Refinement.


HEMM: Holistic Evaluation of Multimodal Foundation Models

Neural Information Processing Systems

Multimodal foundation models that can holistically process text alongside images, video, audio, and other sensory modalities are increasingly used in a variety of real-world applications. However, it is challenging to characterize and study progress in multimodal foundation models, given the range of possible modeling decisions, tasks, and domains. In this paper, we introduce Holistic Evaluation of Multimodal Models (HEMM) to systematically evaluate the capabilities of multimodal foundation models across a set of 3 dimensions: basic skills, information flow, and real-world use cases. Basic multimodal skills are internal abilities required to solve problems, such as learning interactions across modalities, fine-grained alignment, multi-step reasoning, and the ability to handle external knowledge.


Synthcity: a benchmark framework for diverse use cases of tabular synthetic data

Neural Information Processing Systems

Accessible high-quality data is the bread and butter of machine learning research, and the demand for data has exploded as larger and more advanced ML models are built across different domains. Yet, real data often contain sensitive information, are subject to various biases, and are costly to acquire, which compromise their quality and accessibility. Synthetic data have thus emerged as a complement to, sometimes even a replacement for, real data for ML training. However, the landscape of synthetic data research has been fragmented due to the diverse range of data modalities, such as tabular, time series, and images, and the wide array of use cases, including privacy preservation, fairness considerations, and data augmentation. This fragmentation poses practical challenges when comparing and selecting synthetic data generators in for different problem settings. To this end, we develop Synthcity, an open-source Python library that allows researchers and practitioners to perform one-click benchmarking of synthetic data generators across data modalities and use cases. Beyond benchmarking, Synthcity serves as a centralized toolkit for accessing cutting-edge data generators. In addition, Synthcity's flexible plug-in style API makes it easy to incorporate additional data generators into the framework. Using examples of tabular data generation and data augmentation, we illustrate the general applicability of Synthcity, and the insight one can obtain.


Use-Case-Grounded Simulations for Explanation Evaluation

Neural Information Processing Systems

A growing body of research runs human subject evaluations to study whether providing users with explanations of machine learning models can help them with practical real-world use cases. However, running user studies is challenging and costly, and consequently each study typically only evaluates a limited number of different settings, e.g., studies often only evaluate a few arbitrarily selected model explanation methods. To address these challenges and aid user study design, we introduce Simulated Evaluations (SimEvals). SimEvals involve training algorithmic agents that take as input the information content (such as model explanations) that would be presented to the user, to predict answers to the use case of interest. The algorithmic agent's test set accuracy provides a measure of the predictiveness of the information content for the downstream use case. We run a comprehensive evaluation on three real-world use cases (forward simulation, model debugging, and counterfactual reasoning) to demonstrate that SimEvals can effectively identify which explanation methods will help humans for each use case. These results provide evidence that \simevals{} can be used to efficiently screen an important set of user study design decisions, e.g., selecting which explanations should be presented to the user, before running a potentially costly user study.


Continual Learning at the Edge: An Agnostic IIoT Architecture

García-Santaclara, Pablo, Fernández-Castro, Bruno, Díaz-Redondo, Rebeca P., Calvo-Moa, Carlos, Mariño-Bodelón, Henar

arXiv.org Machine Learning

The exponential growth of Internet-connected devices has presented challenges to traditional centralized computing systems due to latency and bandwidth limitations. Edge computing has evolved to address these difficulties by bringing computations closer to the data source. Additionally, traditional machine learning algorithms are not suitable for edge-computing systems, where data usually arrives in a dynamic and continual way. However, incremental learning offers a good solution for these settings. We introduce a new approach that applies the incremental learning philosophy within an edge-computing scenario for the industrial sector with a specific purpose: real time quality control in a manufacturing system. Applying continual learning we reduce the impact of catastrophic forgetting and provide an efficient and effective solution.


Ethics Readiness of Artificial Intelligence: A Practical Evaluation Method

Adomaitis, Laurynas, Israel-Jost, Vincent, Grinbaum, Alexei

arXiv.org Artificial Intelligence

In the governance of emerging technologies, ethical guidance has often relied on so-called soft law instruments--codes of conduct, guidelines, or frameworks--designed to promote responsible behavior without imposing binding legal constraints. This is partly due to the difficulty of imposing harmonized regulations across the EU, especially in a global context characterized by strong reservations expressed by other international actors, e.g. the United States of America, with regard to the regulation of artificial intelligence (AI) that "unduly burdens AI innovation" (Kratsios, Sacks, and Rubio 2025) . Another reason is related to the principle, upheld in several member states such as Germany, that protects scientific freedom by constitutional law. Nevertheless, the recent trajectory of technological regulation in the European Union shows that soft law can evolve into hard law: this has been the case, notably, with the adoption of the AI Act (European Commission 2022; Terpan 2015) .